54 research outputs found

    Component-based Face Detection in Colour Images

    Get PDF
    Abstract: Face detection is an important process in many applications such as face recognition, person identification and tracking, and access control. The technique used for face detection depends on how a face is modelled. In this paper, a face is defined as a skin region and a lips region that meet certain geometrical criteria. Thus, the face detection system has three main components: a skin detection module, a lips detection module, and a face verification module. The Multi-layer perceptron (MLP) neural networks was used for the skin and lips detection modules. In order to test the face detection system, two databases were created. The images in the first database, called In-house, were taken under controlled environment while those in the second database, called WWW, were collected from the World Wide Web and as such have no restriction on lighting, head pose or background. The system achieved a correct detection rate of 87 and 80 percent on the In-house and WWW databases respectivel

    Fusing Facial Features for Face Recognition

    Get PDF
    Face recognition is an important biometric method because of its potential applications in many fields, such as access control, surveillance, and human-computer interaction. In this paper, a face recognition system that fuses the outputs of three face recognition systems based on Gabor jets is presented. The first system uses the magnitude, the second uses the phase, and the third uses the phase-weighted magnitude of the jets. The jets are generated from facial landmarks selected using three selection methods. It was found out that fusing the facial features gives better recognition rate than either facial feature used individually regardless of the landmark selection method

    The Effect of Database Type on Face Recognition Performance for Surveillance Applications

    Get PDF
    Face recognition is one of the most important biometric approaches due to its potential applications in surveillance monitoring and access control. This paper presents a PCA and SVM based face recognition system for surveillance application. A proposed training database selection criteria suitable for surveillance application which consist of 1 mean image per distance class from all the available database sessions is also used for the face recognition system. In this study, the ChokePoint database, specifically the grayscale (PPG) and colored (MPCI) versions of the ChokePoint database, were selected for this work. The objectives of this work is to investigate the effect of the using different training data as well as using different similarity matching method on face recognition for surveillance application. It was found that regardless of the type of databases used, the recognition output pattern on different training data selection criteria was found to be similar. It was also found that regardless of the similarity matching method used, the face recognition system also shows the same recognition performance pattern. The experiment suggests that the proposed training database selection criteria will give similar recognition performance regardless of databases type or face recognition technique used. Overall, the ChokePoint colour database (MPCI) gives better recognition performance than the ChokePoint grayscale database (PPG). Finally, it can be concluded that using 1 mean image per class from all the available database sessions (Case-6) is better compared to using 1 image per class that are randomly selected from all the database sessions (Case-4). Even though a straight comparison between this work proposed system and several published system is not meaningful as different face recognition approaches and experiment criteria are used, nevertheless, this work proposed method performs with 100% recall and reject recognition rate

    Lips detection in the normalised RGB colour scheme

    Get PDF
    Lips detection is used in many applications such as face detection and lips reading. In this paper a method for lips detection in colour images in the normalised RGB colour scheme is presented. A new method, called maximum intensity normalisation, is investigated and compared with the traditional pixel intensity normalisation method. The intensity of the colour image is first normalised using either the pixel or maximum intensity normalisation methods. Then, the skin regions in the image are determined by histogram thresholding the r-b and r-g chrominance components. The skin regions are then thresholded to classify the pixels in the image into lips or non-lips pixels. Two methods for selecting the threshold value where investigated. In the first method, the threshold value is the value that gives the minimum error of both classes. For the second method, the threshold is the value where the error of both classes is the same. Two Asian databases were used. The first was created in-house and the second was collected from the WWW. It was found out that there is no significant difference on the percentage of correct classification between the two threshold selection methods for both databases. However, percentage of correct classification for the in-house database was considerably higher than that of the WWW database

    A Comparison of the YCBCR Color Space with Gray Scale for Face Recognition for Surveillance Applications

    Get PDF
    Face recognition is an important biometric method because of its potential applications in many fields, such as access control and surveillance. In this paper, the performance of the individual channels from the YCBCR colour space on face recognition for surveillance applications is investigated and compared with the performance of the grayscale. In addition, the performance of fusing two or more colour channels is also compared with that of the grayscale. Three cases with a different number of training images per persons were used as a test bed. It was found out that, the grayscale always outperforms the individual channel. However, the fusion of CBxCR with any other channel outperforms the grayscale when three images of the same class from the same database are used for training. Regardless of the cases used, the CBxCR channel always gave the best performance for the individual colour channels. It was found that, in general, increasing the number of fused channels increases the performance of the system. It was also found that the grayscale channel is the better choice for face recognition since it contains better quality edges and visual features which are essential for face recognition

    Performance of machine learning classifiers in distress keywords recognition for audio surveillance applications

    Get PDF
    The ability to recognize distress speech is the essence of an intelligent audio surveillance system. With this ability, the surveillance system can be configured to detect specific distress keywords and launch appropriate actions to prevent unwanted incidents from progressing. This paper aims to find potential distress keywords that the audio surveillance system could recognize. The idea is to use a machine learning classifier as the recognition engine. Five distress keywords: ‘Help’, ‘No’, ‘Oi’, ‘Please’, and ‘Tolong’ were selected to be analyzed. A total of 515 audio signals comprising these five distress keywords were collected and used in the training and testing of 27 classifier models, derived from the Decision Tree, Naïve Bias, Support Vector Machine, K-Nearest Neighbour, Ensemble, and Artificial Neural Network. The features extracted from each audio signal are the Mel-frequency Cepstral Coefficients, while the Principal Component Analysis was applied for feature reduction. The results show that the keyword 'Please' is the most recognized, followed by ‘Help’, ‘Oi’, ‘No’ and ‘Tolong’, respectively. This observation was achieved using the Ensemble Bagged Trees classifier, which can recognize ‘Please’ with 99% accuracy in training and 100% accuracy in testing

    Euclidean space data projection classifier with cartesian genetic programming (CGP)

    Get PDF
    Most evolutionary based classifiers are built based on generated rules sets that categorize the data into respective classes. This research work is a preliminary work which proposes an evolutionary-based classifier using a simplified Cartesian Genetic Programming (CGP) evolutionary algorithm. Instead on using evolutionary generated rule sets, the CGP generates i) a reference coordinate ii) projection functions to project data into a new 3 Dimensional Euclidean space. Subsequently, a distance boundary function of the new projected data to the reference coordinates is applied to classify the data into their respective classes. The evolutionary algorithm is based on a simplified CGP Algorithm using a 1+4 evolutionary strategy. The data projection functions were evolved using CGP for 1000 generations before stopping to extract the best functions. The Classifier was tested using three PROBEN 1 benchmarking datasets which are the PIMA Indians diabetes dataset, Heart Disease dataset and Wisconsin Breast Cancer (WBC) Dataset based on 10 fold cross validation dataset partitioning. Testing results showed that data projection function generated competitive results classification rates: Cancer dataset (97.71%), PIMA Indians dataset (77.92%) and heart disease (85.86%)

    Euclidean Space Data Projection Classifier with Cartesian Genetic Programming (CGP)

    Get PDF
    Most evolutionary based classifiers are built based on generated rules sets that categorize the data into respective classes. This research work is a preliminary work which proposes an evolutionary-based classifier using a simplified Cartesian Genetic Programming (CGP) evolutionary algorithm. Instead on using evolutionary generated rule sets, the CGP generates i) a reference coordinate ii) projection functions to project data into a new 3 Dimensional Euclidean space. Subsequently, a distance boundary function of the new projected data to the reference coordinates is applied to classify the data into their respective classes. The evolutionary algorithm is based on a simplified CGP Algorithm using a 1+4 evolutionary strategy. The data projection functions were evolved using CGP for 1000 generations before stopping to extract the best functions. The Classifier was tested using three PROBEN 1 benchmarking datasets which are the PIMA Indians diabetes dataset, Heart Disease dataset and Wisconsin Breast Cancer (WBC) Dataset based on 10 fold cross validation dataset partitioning. Testing results showed that data projection function generated competitive results classification rates: Cancer dataset (97.71%), PIMA Indians dataset (77.92%) and heart disease (85.86%)
    • …
    corecore